Supplementary material for Collaborative hyperparameter tuning

ثبت نشده
چکیده

In Section 4 of the main paper, we present results on two benchmarks in terms of average ranking, since classification datasets may not be commensurable in terms of raw validation error. For the sake of completeness, we present here results in terms of average meta-test error. Meta-test error is defined slightly differently in our two experiments. We also present a PCA of our data in the MLP experiment.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Collaborative hyperparameter tuning

Hyperparameter learning has traditionally been a manual task because of the limited number of trials. Today’s computing infrastructures allow bigger evaluation budgets, thus opening the way for algorithmic approaches. Recently, surrogate-based optimization was successfully applied to hyperparameter learning for deep belief networks and to WEKA classifiers. The methods combined brute force compu...

متن کامل

Massively Parallel Hyperparameter Tuning

Modern machine learning models are characterized by large hyperparameter search spaces and prohibitively expensive training costs. For such models, we cannot afford to train candidate models sequentially and wait months before finding a suitable hyperparameter configuration. Hence, we introduce the large-scale regime for parallel hyperparameter tuning, where we need to evaluate orders of magnit...

متن کامل

Supplementary Material: Predicting Visual Exemplars of Unseen Classes for Zero-Shot Learning

where θ is an implicit nonlinear mapping based on our kernel. We have dropped the subscript d for aesthetic reasons but readers are reminded that each regressor is trained independently with its own target values (i.e., vcd) and parameters (i.e.,wd). We found that the regression error is not sensitive to λ and set it to 1 in all experiments except for zero-shot to few-shot learning. We jointly ...

متن کامل

YellowFin and the Art of Momentum Tuning

Adaptive Optimization Hyperparameter tuning is a big cost of deep learning. Momentum: a key hyperparameter to SGD and variants. Adaptive methods, e.g. Adam1, don’t tune momentum. YellowFin optimizer • Based on the robustness properties of momentum. • Auto-tuning of momentum and learning rate in SGD. • Closed-loop momentum control for async. training. Experiments ResNet and LSTM YellowFin runs w...

متن کامل

Combination of Hyperband and Bayesian Optimization for Hyperparameter Optimization in Deep Learning

Deep learning has achieved impressive results on many problems. However, it requires high degree of expertise or a lot of experience to tune well the hyperparameters, and such manual tuning process is likely to be biased. Moreover, it is not practical to try out as many different hyperparameter configurations in deep learning as in other machine learning scenarios, because evaluating each singl...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2013